Ingesting data from RDBMS into a Snowflake Data Lake using Snowpark

The Calibo Accelerate platform supports ingesting data into a Snowflake data lake using Snowpark. Data is processed within Snowflake avoiding any data movement. This simplifies the architecture and effectively improves the performance of the ingestion pipeline.

Currently the Calibo Accelerate platform supports the following RDBMS data sources for data ingestion into Snowflake data lake using Snowpark:

  • MySQL

  • PostgreSQL

  • Oracle

To create a data integration job using Snowpark

  1. Sign in to the Calibo Accelerate platform and navigate to Products.

  2. Select a product and feature. Click the Develop stage of the feature and navigate to Data Pipeline Studio.

  3. Create a data pipeline with the following stages:

    Data source (MySQL) > Data Integration (Snowpark) > Data Lake (Snowflake)

    Data ingestion using Snowpark

  4. Add MySQL in the data source stage and Snowflake in the data lake stage and configure both the nodes. To know more about configuring the source node, see Data Sources.

  5. In the data integration node, add Snowpark and do the following:

    1. Provide a technology title.

    2. Select a Snowpark instance.

      Note:

      The Snowpark integration node must be configured using the same Snowflake account that is used in the data lake stage.

    3. Click Save.

  6. Click the Snowpark node and click Create Job.

    Create a templatized ingestion job using Snowpark

    1. Complete the following steps to create the Snowpark integration job:

Related Topics Link IconRecommended Topics What's next? Snowpark Custom Transformation Job